( ) = Y ˆ. Calibration Definition A model is calibrated if its predictions are right on average: ave(response Predicted value) = Predicted value.

Similar documents
Section 3.4: Diagnostics and Transformations. Jared S. Murray The University of Texas at Austin McCombs School of Business

Further Maths Notes. Common Mistakes. Read the bold words in the exam! Always check data entry. Write equations in terms of variables

Graphical Analysis of Data using Microsoft Excel [2016 Version]

Using Excel for Graphical Analysis of Data

Section 3.2: Multiple Linear Regression II. Jared S. Murray The University of Texas at Austin McCombs School of Business

Instruction on JMP IN of Chapter 19

Multiple Regression White paper

EC121 Mathematical Techniques A Revision Notes

Stat 401 B Lecture 26

Algebra 2 Semester 1 (#2221)

Using Excel for Graphical Analysis of Data

Building Better Parametric Cost Models

Section 2.3: Simple Linear Regression: Predictions and Inference

Your Name: Section: INTRODUCTION TO STATISTICAL REASONING Computer Lab #4 Scatterplots and Regression

SLStats.notebook. January 12, Statistics:

STAT 705 Introduction to generalized additive models

Multiple Linear Regression

MATH 115: Review for Chapter 1

CCNY Math Review Chapter 2: Functions

Lecture 24: Generalized Additive Models Stat 704: Data Analysis I, Fall 2010

Two-Stage Least Squares

HW 10 STAT 472, Spring 2018

Machine Learning and Bioinformatics 機器學習與生物資訊學

2 Computation with Floating-Point Numbers

Four equations are necessary to evaluate these coefficients. Eqn

STA 570 Spring Lecture 5 Tuesday, Feb 1

Course Number 432/433 Title Algebra II (A & B) H Grade # of Days 120

AM205: lecture 2. 1 These have been shifted to MD 323 for the rest of the semester.

Modesto City Schools. Secondary Math I. Module 1 Extra Help & Examples. Compiled by: Rubalcava, Christina

YEAR 12 Trial Exam Paper FURTHER MATHEMATICS. Written examination 1. Worked solutions

VCEasy VISUAL FURTHER MATHS. Overview

Prepare a stem-and-leaf graph for the following data. In your final display, you should arrange the leaves for each stem in increasing order.

Rational functions, like rational numbers, will involve a fraction. We will discuss rational functions in the form:

Nonparametric Approaches to Regression

STANDARDS OF LEARNING CONTENT REVIEW NOTES ALGEBRA I. 4 th Nine Weeks,

Interactive Math Glossary Terms and Definitions

HW 10 STAT 672, Summer 2018

ALGEBRA 1 NOTES. Quarter 3. Name: Block

DESIGN OF EXPERIMENTS and ROBUST DESIGN

6th Grade P-AP Math Algebra

Chapter 4: Analyzing Bivariate Data with Fathom

ES-2 Lecture: Fitting models to data

Section 2.1: Intro to Simple Linear Regression & Least Squares

Voluntary State Curriculum Algebra II

The problem we have now is called variable selection or perhaps model selection. There are several objectives.

STAT 2607 REVIEW PROBLEMS Word problems must be answered in words of the problem.

IQR = number. summary: largest. = 2. Upper half: Q3 =

Linear Regression in two variables (2-3 students per group)

Goals of the Lecture. SOC6078 Advanced Statistics: 9. Generalized Additive Models. Limitations of the Multiple Nonparametric Models (2)

1. Answer: x or x. Explanation Set up the two equations, then solve each equation. x. Check

Experiment 1 CH Fall 2004 INTRODUCTION TO SPREADSHEETS

THIS IS NOT REPRESNTATIVE OF CURRENT CLASS MATERIAL. STOR 455 Midterm 1 September 28, 2010

MAT 003 Brian Killough s Instructor Notes Saint Leo University

Math 1: Solutions to Written Homework 1 Due Friday, October 3, 2008

Understanding and Comparing Distributions. Chapter 4

Section 1.5 Transformation of Functions

Mathematics Scope & Sequence Grade 8 Revised: June 2015

VW 1LQH :HHNV 7KH VWXGHQW LV H[SHFWHG WR

Grade 9 Math Terminology

Statistical Good Practice Guidelines. 1. Introduction. Contents. SSC home Using Excel for Statistics - Tips and Warnings

1. Assumptions. 1. Introduction. 2. Terminology

Section 1.5 Transformation of Functions

IE 361 Exam 1 October 2005 Prof. Vardeman Give Give Does Explain What Answer explain

Six Weeks:

The first thing we ll need is some numbers. I m going to use the set of times and drug concentration levels in a patient s bloodstream given below.

Things to Know for the Algebra I Regents

Direct Variations DIRECT AND INVERSE VARIATIONS 19. Name

Design and Analysis of Algorithms Prof. Madhavan Mukund Chennai Mathematical Institute

Data can be in the form of numbers, words, measurements, observations or even just descriptions of things.

Important Things to Remember on the SOL

An introduction to plotting data

Tips and Guidance for Analyzing Data. Executive Summary

STANDARDS OF LEARNING CONTENT REVIEW NOTES. ALGEBRA I Part II. 3 rd Nine Weeks,

2 Computation with Floating-Point Numbers

Multicollinearity and Validation CIVL 7012/8012

NOTES Linear Equations

Unit 2-2: Writing and Graphing Quadratics NOTE PACKET. 12. I can use the discriminant to determine the number and type of solutions/zeros.

CHAPTER 3 AN OVERVIEW OF DESIGN OF EXPERIMENTS AND RESPONSE SURFACE METHODOLOGY

2D Spline Curves. CS 4620 Lecture 18

The Time Series Forecasting System Charles Hallahan, Economic Research Service/USDA, Washington, DC

Scatterplot: The Bridge from Correlation to Regression

ALGEBRA II A CURRICULUM OUTLINE

3 Nonlinear Regression

Unit Essential Questions: Does it matter which form of a linear equation that you use?

Glossary Common Core Curriculum Maps Math/Grade 6 Grade 8

Name Course Days/Start Time

Lecture 1: Statistical Reasoning 2. Lecture 1. Simple Regression, An Overview, and Simple Linear Regression

Spatial Patterns Point Pattern Analysis Geographic Patterns in Areal Data

CPSC 340: Machine Learning and Data Mining. More Regularization Fall 2017

Non-Linearity of Scorecard Log-Odds

2.1 Solutions to Exercises

Lesson 1: Analyzing Quadratic Functions

THE L.L. THURSTONE PSYCHOMETRIC LABORATORY UNIVERSITY OF NORTH CAROLINA. Forrest W. Young & Carla M. Bann

Summer Review for Students Entering Pre-Calculus with Trigonometry. TI-84 Plus Graphing Calculator is required for this course.

HPISD Eighth Grade Math

2.1. Rectangular Coordinates and Graphs. 2.1 Rectangular Coordinates and Graphs 2.2 Circles 2.3 Functions 2.4 Linear Functions. Graphs and Functions

Today is the last day to register for CU Succeed account AND claim your account. Tuesday is the last day to register for my class

Preview Notes. Systems of Equations. Linear Functions. Let y = y. Solve for x then solve for y

Chapter 6: DESCRIPTIVE STATISTICS

Table of Laplace Transforms

Transcription:

Calibration OVERVIEW... 2 INTRODUCTION... 2 CALIBRATION... 3 ANOTHER REASON FOR CALIBRATION... 4 CHECKING THE CALIBRATION OF A REGRESSION... 5 CALIBRATION IN SIMPLE REGRESSION (DISPLAY.JMP)... 5 TESTING FOR A LACK OF CALIBRATION... 8 CALIBRATION PLOT...1 CHECKING THE CALIBRATION...12 CHECKING CALIBRATION IN MULTIPLE REGRESSION...14 CALIBRATING A MODEL...17 PREDICTING WITH A SMOOTHING SPLINE...18 OTHER FORMS OF CALIBRATION...19

Overview This lecture introduces calibration along with methods for checking whether a model is calibrated. Calibration Calibration is a fundamental property of any predictive model. If a model is not calibrated its predictions may systematically lead to poor decisions. Test for calibration uses the plot of Y on the fitted values. It is similar to checking for a nonlinear pattern in the scatterplot of Y versus X with a transformation. New tool Smoothing splines fit smooth trends that may not be linear. Calibrating a model using smoothing splines improves its predictions of the response, albeit without offering more in the way of explanation. Introduction Better models produce better predictions in several ways: The better the model, the more closely its predictions track the average of the response. The better the model, the more precisely its predictions match the response (e.g., smaller prediction intervals). The better the model, the more likely it is that the distribution of the prediction errors are normally distributed (as assumed by the MRM). Consequences of systematic errors The first aspect, to be right on average, is critical. Unless predictions from a model are right on average, the model cannot be used for economic decisions. Calibration is about being right on average. Statistics 471/71 4-2 Spring 21

High R 2 calibrated. Calibration is neither a consequence nor precursor to a large R 2. A model with R 2 =.5 may be calibrated, and a model with R 2 =.999 need not be calibrated. In simple regression, calibration is related to the choice of transformations that capture nonlinear transformations. Calibration Definition A model is calibrated if its predictions are right on average: ave(response Predicted value) = Predicted value. ( ) = Y ˆ E Y ˆ Y Calibration is a basic property of any predictor, whether the prediction results from the equation of a regression or a gypsy s forecast from a crystal ball. A regression should be calibrated before we use its predictions. Calibration in other situations Physician s opinions on risk of an illness Weather forecasts Political commentary on the state of financial markets Example A bank has made a collection of loans. It expects to earn $1 interest on average from these loans each month. If this claim is calibrated, then the average interest earned over following months will be $1. Statistics 471/71 4-3 Spring 21

If the bank earns $ half of the months and $2 during the other months, then the estimate is calibrated. (Not very precise, but it would be calibrated.) If the bank earns $15 every month, then the estimate is not calibrated (even though it is arguably a better prediction). Role in demand models (Module 6) Getting the mean level of demand correct is the key to making a profitable decision. Calibration implies that predicted demand is indeed the mean demand µ under the conditions described by the explanatory variables. Example: If you collect days on which we predict to sell 15 items, the average sales on these days is 15. Calibration does not imply small variance or a good estimate of σ. It only means that we are right on average. Another Reason for Calibration A gift to competitors If a model is not calibrated, a rival can improve its predictions without having to model the response. Rivals have an easy way to beat your estimates. Example You spend a lot of effort to develop a model that predicts the credit risk of customers. If these predictions are not calibrated, a rival with no domain knowledge can improve these predictions, obtaining a better estimate of the credit risk and identify the more profitable customers. Do you really want to miss out? Statistics 471/71 4-4 Spring 21

Checking the Calibration of a Regression Basic procedure is familiar No new tools are absolutely necessary: you ve been checking for calibration all along, but did not use this terminology. Checking for calibration uses methods that you have already seen in other contexts. Calibration in simple regression A simple regression is calibrated if the fit of the equation tracks the average of the response. Often, to get a calibrated model, you transform one or both of the variables (e.g., using logs or reciprocals) Calibration in multiple regression The calibration plot is part of the standard output produced by JMP that describes the fit of a multiple regression. Calibration in Simple Regression (display.jmp) Relationship between promotion and sales 1 How much of a new beverage product should be put on display in a chain of liquor stores? 1 This example begins in BAUR on page 12 and shows up several times in that casebook. This example illustrates the problem of resource allocation because to display more of one item requires showing less of items. The resource is the limited shelf space in stores. Statistics 471/71 4-5 Spring 21

Scatterplot and linear fit Sales 45 4 35 3 25 2 15 1 5 1 2 3 4 5 6 7 8 Display Feet Sales = 93 + 39.8 Display Feet The linear fit misses the average amount sold for each amount of shelf space devoted to its display. The objective is to identify a smooth curve that captures the relationship between the display feet and the amount sold at the stores. Smoothing spline We can show the average sales for each number of display feet by adding a smoothing spline to the plot. 2 A spline is a smooth curve that connects the average value of Y over a subset of values of identified by a subset of adjacent values of X. The smoothness of the spline depends on the data and can be controlled by a slider in the software. The slider controls the size of the averaged subset by controlling the range of values of X. Example of spline In this example, a smoothing spline joins the mean sales for each amount sold with a smooth curve. 3 2 In the Fit Y by X view of the data, choose the Fit Spline option from the tools revealed by the red triangle in the upper left corner of the window. Pick the flexible option. Statistics 471/71 4-6 Spring 21

Sales 45 4 35 3 25 2 15 1 5 1 2 3 4 5 6 7 8 Display Feet With replications (several stores display the same amount; these data have several y s for each x), the data form columns of dots in the plot. If the range of adjacent values of X is small enough (less than 1) then the spline joins the means of these columns. 4 Without replications, the choice of a spline is less obvious and can require a bit of practice (i.e., guesswork). Compare the spline to the linear fit With one foot on display, the linear fit seems too large. The mean is about $5. With two feet on display, the linear fit seems too small. The mean of sales is around $2. The linear fit explains R 2 = 71% of the variation. A spline always does as least as well as the linear fit. For the spline, R 2 = 86%. The situation resembles multiple regression: R 2 grows as we add predictors. In a sense, a spline adds predictors as well (but JMP does not tell us how many). 3 The spline tool is easy to control. Use the slider to vary the smoothness of the curve shown in the plot. See BAUR casebook, p. 13, for further discussion of splines and these data. 4 Replications (several cases with matching values of all of the Xs) are a great asset if you have them. Usually, they only show up in simple regression because there are too many combinations of predictors in multiple regression. Because this dataset has replications, we can use JMP s Lack of Fit test to check for calibration. See the BAUR casebook, page 247. Statistics 471/71 4-7 Spring 21

Testing for a Lack of Calibration Objective Does the spline really fit the data better that the line (able to predict new observations more accurately)? Does the spline capitalize on random, idiosyncratic features of the data that won t show up in new observations? Approximate spline with polynomial Approximate the spline by fitting a polynomial that roughly matches its fit to the data. We can test the polynomial fit to see if its better than the linear fit. Fit a polynomial model 5 This plot superimposes the fit of a 5 th degree polynomial ˆ y = b + b 1 X + b 2 X 2 + b 3 X 3 + b 4 X 4 + b 5 X 5 The output centers the predictor (subtracting the mean 4.4 to reduce the collinearity). We only need R 2. Sales 45 4 35 3 25 2 15 1 5 1 2 3 4 5 6 7 8 Display Feet RSquare.8559 Root Mean Square Error 38.22968 Term Estimate Std Error t Ratio Prob> t Intercept 19.52749 7.2328 1.56.1266 Display Feet 37.9317 15.51394 2.45.189 (Display Feet-4.4426)^2 12.189222 9.94144 1.34.1875 (Display Feet-4.4426)^3-2.36853 6.6657 -.39.6982 (Display Feet-4.4426)^4-2.7732 1.23635-1.67.116 (Display Feet-4.4426)^5.18778.5666.19.847 5 To add a polynomial to a scatterplot, click the red triangle in the upper left corner of the output window and pick Fit Polynomial. To match the spline often requires 5 or 6 powers. Statistics 471/71 4-8 Spring 21

Partial F-test Do not use the t-statistics to assess the polynomial because of collinearity (even with centering). We re not interpreting the coefficients; all we want to learn is whether R 2 significantly improved. The partial F-test does what we need. It finds a very significant increase for the 4 added predictors. The degrees-of-freedom divisor in the numerator of the ratio is 4 (not 5) because the polynomial adds 4 more predictors. The d.f. for the denominator of F is n 2 (for the original fit) 4 (added by the spline) (.8559.712) /4 F = (1.8559) /(47 6) = 41 4.1439.1441 =1.2 Conclude: The linear regression is not calibrated. 6 Calibrating simple regression Transform one or both variables to capture the pattern in the means and obtain a more interpretable model that fits almost as well. Substantively, model the diminishing returns to scale using a log of the number of display feet as a predictor. The R 2 of the fit using log(display feet) is 81.5% and the partial F improvement offered by the polynomial is no longer significant (F = 2.9). 7 6 The improvement is statistically significant even though none of the t-stats for the added coefficients is significant. There s too much collinearity. When you add several terms at once, use the partial F to judge the increase. 7 F= ((.8559-.815)/4)/((1-.8559)/41) = 2.9 Statistics 471/71 4-9 Spring 21

Sales 45 4 35 3 25 2 15 1 5 1 2 3 4 5 6 7 8 Display Feet Sales = 83.56 + 138.62 Log(Display Feet) Calibration Plot Generalize to multiple regression A calibration plot offers an equivalent test of calibration, one that generalizes to multiple regression. We need this generalization to test for lack of calibration more convenient in models with more than one predictor. Definition A calibration plot is a scatterplot of the actual values of the response on the y-axis and the predicted values on the x-axis. (This plot is part of the default output produced by Fit Model.) Sales Actual 45 4 35 3 25 2 15 1 5 5 1 15 2 25 3 35 4 45 Sales Predicted P<.1 RSq=.71 RMSE=51.591 We need to make our own calibration plot in order to test for calibration. Save the prediction formula (or the predicted/fitted values) from the simple regression, then scatterplot the actual values on the predicted values. Statistics 471/71 4-1 Spring 21

45 4 35 3 Sales 25 2 15 1 5 5 1 15 2 25 3 35 4 45 Pred Formula Sales The slope of this fit must be b = 1 and the intercept must be b 1 =. The R 2 and RMSE match those of the previous regression. Why is this true? Linear Fit 8 Sales = + 1 Pred Formula Sales Summary of Fit RSquare.712 Root Mean Square Error 51.591 Observations (or Sum Wgts) 47 Parameter Estimates Term Estimate Std Error t Ratio Prob> t Intercept 26.51. 1. Pred Formula Sales 1.9 1.55 <.1 Hint: Recall that R 2 in regression is the square of the correlation between the response and the predicted response. 8 JMP sometimes uses scientific notation for numbers very close to zero, something like 1.e- 13, which translates to 1 1-13. The difference from zero in this output is due to round-off errors in the underlying numerical calculations; computers use binary arithmetic that does not represent exactly every number that you write down. Statistics 471/71 4-11 Spring 21

Checking the Calibration Using the calibration plot We can check the calibration of any regression from the calibration plot. Example This summary of what happens when we add the 5 th order polynomial for the display example. 9 The coefficients of the polynomial differ from those in the prior fit, but R 2 and RMSE of the fit are the same. Caution: Be careful computing the partial F test. We want the degrees of freedom as though we added these terms to the original model as done previously, not started over. Hence, we reach the same conclusion from the calibration plot as when working with the scatterplot of Y on X: the linear model not calibrated. Sales 45 4 35 3 25 2 15 1 5 5 1 15 2 25 3 35 4 45 Pred Formula Sales The summary of the fit of this model is on the next page. 9 Why use a 5 th order polynomial, you might ask? It s arbitrary, but seems to match the fit of a smoothing spline so long as we stay away from the edges of the plot. Statistics 471/71 4-12 Spring 21

Summary of Fit RSquare.8559 Root Mean Square Error 38.2297 Mean of Response 268.13 Observations (or Sum Wgts) 47. Parameter Estimates Term Estimate Std Error t Ratio Prob> t Intercept 2.765257 16.1972.2.8459 Pred Formula Sales.954111.39224 2.45.189 (Pred Formula Sales-268.13)^2.77119.5754 1.34.1875 (Pred Formula Sales-268.13)^3 -.38.97 -.39.6982 (Pred Formula Sales-268.13)^4-8.29e-7 4.9e-7-1.67.116 (Pred Formula Sales-268.13)^5 1.1e-9 5.6e-9.19.847 Warning: Never extrapolate a high-order polynomial The fit of a polynomial of this degree goes wild when extrapolated beyond the data, even though it seems reasonable within the observations. The spline is much more useful for this purpose. When to check for calibration? Calibration is a diagnostic, something done once you feel you have a reasonable model and are ready to start checking residuals. Statistics 471/71 4-13 Spring 21

Checking Calibration in Multiple Regression Procedure Use the calibration plot of y on the fitted values y ˆ (a.k.a., predicted values) from the regression. Test the calibration by seeing whether a polynomial fit in the plot of y on y ˆ significantly improves upon the linear fit that summarizes the R 2 and RMSE of the regression. Tricky part The degrees of freedom for the associated partial F test depend on the number of slopes in the original multiple regression. Thankfully, this adjustment has little effect unless you have a small sample. Example data Model of housing prices, from the analysis of the effects of pollution on residential home prices in the 197 s in the Boston area (introduced in Module 3). In the examples of interactions, we observed curvature in the residuals of the models. That s a lack of calibration. Illustrative multiple regression This regression with 5 predictors explains almost 7% of the variation in housing values. All of the predictors are statistically significant (if we accept the MRM) with tiny p- values. 1 The model lacks interactions; every predictor is assumed to have a linear partial association with the prices, with a fixed slope regardless of other explanatory variables. 1 This model is part of the JMP data file; run the Uncalibrated regression script. Statistics 471/71 4-14 Spring 21

Term Summary of Fit RSquare.689 Root Mean Square Error 4.951 Mean of Response 22.248 Observations 5 Parameter Estimates Estimate Std Error t Ratio Prob> t Intercept 72.1 3.11 23.18 <.1 NOx -2.83 3.25-6.41 <.1 Distance -1.55.19-8.18 <.1 Lower class -.74.4-17.66 <.1 Pupil/Teacher -1.28.12-1.88 <.1 Zoning.5.1 3.48.6 Calibration plot We find curvature in the plot of the housing values on the fitted values from the regression. As happens often, a lack of calibration is most evident at the extremes with the smallest and largest fitted values. 5 4 Value 3 2 1 1 2 3 4 Pred Formula Value RSquare.689 Root Mean Square Error 4.93 Term Estimate Std Error t Ratio Prob> t Intercept -7.5e-14.76. 1. Pred Formula Value 1.3 33.18 <.1 The fitted slope and intercept are 1 and, with R 2 and RMSE matching those in the multiple regression. Testing for a lack of calibration Statistics 471/71 4-15 Spring 21

Use the procedure that we applied to simple regression, starting from a scatterplot of y on y ˆ. The fitted 5 th order polynomial (or spline) shows that average housing values are larger than the fitted values at both sides of the plot. The fitted values are too small for both inexpensive and expensive housing. The effect for inexpensive tracts (left) seems less noticeable than for the expensive tracts (right). 5 4 Value 3 2 1 1 2 3 4 5 Pred Formula Value The partial F-test for calibration shows that there s room to improve the model. 11 Summary of Fit RSquare.742 Root Mean Square Error 4.56 Observations (or Sum Wgts) 5 (.742.689) /4 F = (1.742) /(5 6 4) = 49 4.53.258 = 25.2 11 The degrees of freedom in the F are 4 (for the 4 nonlinear terms in the polynomial) and n 6 4. (6 for the intercept and slopes in the original fit plus 4 for the non-linear terms). Statistics 471/71 4-16 Spring 21

Calibrating a Model What should be done when a model is not calibrated? Simple regression. Ideally, find a substantively motivated transformation, such as a log, that captures the curvature. Data mining is no substitute for knowing the substance of the problem. Multiple regression. Again, find the right transformation or a missing predictor. This can be hard to do, but some methods can find these (and indeed work for this data set and model). If you only need predictions, calibrate the fit. (a) Use predictions from the polynomial used to test for calibration, or better yet (b) Use a spline that matches the polynomial fit in the test for calibration. The spline avoids the edge effects that make polynomials go wild when extrapolating. Example The matching spline (lambda=238.43) has similar R 2 to the fit of the polynomial model used in the test (74.2%), but the spline controls the behavior at the edges of the plot. 5 4 Value 3 2 1 R-Square.748328 Sum of Squares Error 9784.732 1 2 3 4 5 Pred Formula Value Statistics 471/71 4-17 Spring 21

Predicting with a Smoothing Spline Use this three-step process: 1. Save predictions y ˆ from the original, uncalibrated multiple regression (used to test the calibration). 2. Match a smoothing spline to the polynomial in the calibration plot (as done above). 3. Save the prediction formula from this smoothing spline as a column in the data table. 4. Insert values of y ˆ from the regression into the formula for the smoothing spline to get the final prediction. Example Predict the first row. This row is missing the response column and matches the predictors held in row #26, a case that the fitted regression under-predicts. Predicted value from multiple regression = 39. Predict the response using the spline fit for cases with this predicted value. If you have saved the formula for the spline fit, this calculation happens automagically. Predicted value from the spline = 49.5 The actual value for the matching case in row #26 is 5. The spline gives a much better estimate! The use of a spline after a regression fixes the calibration, but avoids any attempt at explanation. That s a weakness. Knowing why the model under-predicts housing values at either extreme might be important. Statistics 471/71 4-18 Spring 21

Other Forms of Calibration When a model predicts the response right, on average, we might ought to say that the model is calibrated to first order. Higher-order calibration matters as well: (a) Second-order calibrated: the variation around every prediction is a known or correctly estimated (including capturing changes in error variation). (b) Strongly calibrated: the distribution of the prediction errors is known or correctly estimated (such as with a normal distribution). We typically check these as part of the usual regression diagnostics, after we get the right equation for the mean value. That s the message here as well: Get the predicted value right on average, then worry about the other attributes of your prediction. Statistics 471/71 4-19 Spring 21